Goto

Collaborating Authors

 online community


Asking For It: Question-Answering for Predicting Rule Infractions in Online Content Moderation

Samory, Mattia, Pamfile, Diana, To, Andrew, Phadke, Shruti

arXiv.org Artificial Intelligence

Online communities rely on a mix of platform policies and community-authored rules to define acceptable behavior and maintain order. However, these rules vary widely across communities, evolve over time, and are enforced inconsistently, posing challenges for transparency, governance, and automation. In this paper, we model the relationship between rules and their enforcement at scale, introducing ModQ, a novel question-answering framework for rule-sensitive content moderation. Unlike prior classification or generation-based approaches, ModQ conditions on the full set of community rules at inference time and identifies which rule best applies to a given comment. We implement two model variants - extractive and multiple-choice QA - and train them on large-scale datasets from Reddit and Lemmy, the latter of which we construct from publicly available moderation logs and rule descriptions. Both models outperform state-of-the-art baselines in identifying moderation-relevant rule violations, while remaining lightweight and interpretable. Notably, ModQ models generalize effectively to unseen communities and rules, supporting low-resource moderation settings and dynamic governance environments.


Reading the post-riot posts: how we traced far-right radicalisation across 51,000 Facebook messages

The Guardian

Jail sentences for those who made posts about the UK riots in summer 2024 have become a flashpont for online criticism. Jail sentences for those who made posts about the UK riots in summer 2024 have become a flashpont for online criticism. More than 1,100 people have been charged in connection to the summer 2024 riots. A small number of them were charged for offences related to their online activity. Their jail sentences - which ranged from 12 weeks to seven years - became a flashpoint for online criticism.


Exploring Self-Identified Counseling Expertise in Online Support Forums

Lahnala, Allison, Zhao, Yuntian, Welch, Charles, Kummerfeld, Jonathan K., An, Lawrence, Resnicow, Kenneth, Mihalcea, Rada, Pérez-Rosas, Verónica

arXiv.org Artificial Intelligence

A growing number of people engage in online health forums, making it important to understand the quality of the advice they receive. In this paper, we explore the role of expertise in responses provided to help-seeking posts regarding mental health. We study the differences between (1) interactions with peers; and (2) interactions with self-identified mental health professionals. First, we show that a classifier can distinguish between these two groups, indicating that their language use does in fact differ. To understand this difference, we perform several analyses addressing engagement aspects, including whether their comments engage the support-seeker further as well as linguistic aspects, such as dominant language and linguistic style matching. Our work contributes toward the developing efforts of understanding how health experts engage with health information- and support-seekers in social networks. More broadly, it is a step toward a deeper understanding of the styles of interactions that cultivate supportive engagement in online communities.


PrompTrend: Continuous Community-Driven Vulnerability Discovery and Assessment for Large Language Models

Gasmi, Tarek, Guesmi, Ramzi, Aloui, Mootez, Bennaceur, Jihene

arXiv.org Artificial Intelligence

Static benchmarks fail to capture LLM vulnerabilities emerging through community experimentation in online forums. We present PrompTrend, a system that collects vulnerability data across platforms and evaluates them using multidimensional scoring, with an architecture designed for scalable monitoring. Cross-sectional analysis of 198 vulnerabilities collected from online communities over a five-month period (January-May 2025) and tested on nine commercial models reveals that advanced capabilities correlate with increased vulnerability in some architectures, psychological attacks significantly outperform technical exploits, and platform dynamics shape attack effectiveness with measurable model-specific patterns. The PrompTrend Vulnerability Assessment Framework achieves 78% classification accuracy while revealing limited cross-model transferability, demonstrating that effective LLM security requires comprehensive socio-technical monitoring beyond traditional periodic assessment. Our findings challenge the assumption that capability advancement improves security and establish community-driven psychological manipulation as the dominant threat vector for current language models.


K/DA: Automated Data Generation Pipeline for Detoxifying Implicitly Offensive Language in Korean

Jeon, Minkyeong, Jeong, Hyemin, Kim, Yerang, Kim, Jiyoung, Cho, Jae Hyeon, Lee, Byung-Jun

arXiv.org Artificial Intelligence

Language detoxification involves removing toxicity from offensive language. While a neutral-toxic paired dataset provides a straightforward approach for training detoxification models, creating such datasets presents several challenges: i) the need for human annotation to build paired data, and ii) the rapid evolution of offensive terms, rendering static datasets quickly outdated. To tackle these challenges, we introduce an automated paired data generation pipeline, called K/DA. This pipeline is designed to generate offensive language with implicit offensiveness and trend-aligned slang, making the resulting dataset suitable for detoxification model training. We demonstrate that the dataset generated by K/DA exhibits high pair consistency and greater implicit offensiveness compared to existing Korean datasets, and also demonstrates applicability to other languages. Furthermore, it enables effective training of a high-performing detoxification model with simple instruction fine-tuning.


BTPD: A Multilingual Hand-curated Dataset of Bengali Transnational Political Discourse Across Online Communities

Das, Dipto, Ahmed, Syed Ishtiaque, Guha, Shion

arXiv.org Artificial Intelligence

Understanding political discourse in online spaces is crucial for analyzing public opinion and ideological polarization. While social computing and computational linguistics have explored such discussions in English, such research efforts are significantly limited in major yet under-resourced languages like Bengali due to the unavailability of datasets. In this paper, we present a multilingual dataset of Bengali transnational political discourse (BTPD) collected from three online platforms, each representing distinct community structures and interaction dynamics. Besides describing how we hand-curated the dataset through community-informed keyword-based retrieval, this paper also provides a general overview of its topics and multilingual content.


Dataset reveals how Reddit communities are adapting to AI

AIHub

Researchers at Cornell Tech have released a dataset extracted from more than 300,000 public Reddit communities, and a report detailing how Reddit communities are changing their policies to address a surge in AI-generated content. The team collected metadata and community rules from the online communities, known as subreddits, during two periods in July 2023 and November 2024. The researchers will present a paper with their findings at the Association of Computing Machinery's CHI conference on Human Factors in Computing Systems being held April 26 to May 1 in Yokohama, Japan. One of the researchers' most striking discoveries is the rapid increase in subreddits with rules governing AI use. According to the research, the number of subreddits with AI rules more than doubled in 16 months, from July 2023 to November 2024. "This is important because it demonstrates that AI concern is spreading in these communities.


Linguistic Comparison of AI- and Human-Written Responses to Online Mental Health Queries

Saha, Koustuv, Jain, Yoshee, De Choudhury, Munmun

arXiv.org Artificial Intelligence

The ubiquity and widespread use of digital and online technologies have transformed mental health support, with online mental health communities (OMHCs) providing safe spaces for peer support. More recently, generative AI and large language models (LLMs) have introduced new possibilities for scalable, around-the-clock mental health assistance that could potentially augment and supplement the capabilities of OMHCs. Although genAI shows promise in delivering immediate and personalized responses, their effectiveness in replicating the nuanced, experience-based support of human peers remains an open question. In this study, we harnessed 24,114 posts and 138,758 online community (OC) responses from 55 OMHCs on Reddit. We prompted several state-of-the-art LLMs (GPT-4-Turbo, Llama-3, and Mistral-7B) with these posts, and compared their (AI) responses to human-written (OC) responses based on a variety of linguistic measures across psycholinguistics and lexico-semantics. Our findings revealed that AI responses are more verbose, readable, and analytically structured, but lack linguistic diversity and personal narratives inherent in human-human interactions. Through a qualitative examination, we found validation as well as complementary insights into the nature of AI responses, such as its neutrality of stance and the absence of seeking back-and-forth clarifications. We discuss the ethical and practical implications of integrating generative AI into OMHCs, advocating for frameworks that balance AI's scalability and timeliness with the irreplaceable authenticity, social interactiveness, and expertise of human connections that form the ethos of online support communities.


Socially Constructed Treatment Plans: Analyzing Online Peer Interactions to Understand How Patients Navigate Complex Medical Conditions

Basak, Madhusudan, Sharif, Omar, Hulsey, Jessica, Saunders, Elizabeth C., Goodman, Daisy J., Archibald, Luke J., Preum, Sarah M.

arXiv.org Artificial Intelligence

When faced with complex and uncertain medical conditions (e.g., cancer, mental health conditions, recovery from substance dependency), millions of patients seek online peer support. In this study, we leverage content analysis of online discourse and ethnographic studies with clinicians and patient representatives to characterize how treatment plans for complex conditions are "socially constructed." Specifically, we ground online conversation on medication-assisted recovery treatment to medication guidelines and subsequently surface when and why people deviate from the clinical guidelines. We characterize the implications and effectiveness of socially constructed treatment plans through in-depth interviews with clinical experts. Finally, given the enthusiasm around AI-powered solutions for patient communication, we investigate whether and how socially constructed treatment-related knowledge is reflected in a state-of-the-art large language model (LLM). Leveraging a novel mixed-method approach, this study highlights critical research directions for patient-centered communication in online health communities.


Video games can't escape their role in the radicalisation of young men Keith Stuart

The Guardian

There is a lot of attention on young men and toxic masculinity at the moment. The devastating Netflix drama Adolescence, about a 13-year-old boy accused of murdering a girl after being radicalised by the online manosphere, has drawn attention to the problem through the sheer force of its brilliant writing and a blistering lead performance from teenager Owen Cooper. Recently, former England football manager Gareth Southgate gave a speech about the state of boyhood in the UK, specifically about how young men, lacking moral mentors, are turning to gambling and video gaming, thereby disconnecting from society and immersing themselves in predominantly male online communities where misogyny and racism are often rife. There has been some kickback in the gaming press to the idea that games have provided a less-than-ideal environment for boys, but even those of us who have played and enjoyed games all our lives need to face up to the fact that gaming forums, message boards, streaming platforms and social media groups are awash with disturbing hate speech and violent rhetoric. Honestly, we have known this for a while.